sylvester flow
Supplementary Material for: The Convolution Exponential and Generalized Sylvester Flows
The inverse of Sylvester flows can be easily computed using a fixed point iteration. The setup is identical to section C.1, where a single subflow is now either a residual block or a convolutional Sylvester flow transformation, with a leading actnorm layer [ Results are obtained by running models a single after random weight initialization. Additionally, the gated convolutions are replaced by denseblock layers.
- Europe > Netherlands > North Holland > Amsterdam (0.05)
- North America > Canada (0.04)
- North America > Canada > British Columbia > Metro Vancouver Regional District > Vancouver (0.15)
- Europe > Netherlands > North Holland > Amsterdam (0.05)
- North America > United States > California > Los Angeles County > Long Beach (0.05)
- (5 more...)
The Convolution Exponential and Generalized Sylvester Flows
This paper introduces a new method to build linear flows, by taking the exponential of a linear transformation. This linear transformation does not need to be invertible itself, and the exponential has the following desirable properties: it is guaranteed to be invertible, its inverse is straightforward to compute and the log Jacobian determinant is equal to the trace of the linear transformation. An important insight is that the exponential can be computed implicitly, which allows the use of convolutional layers. Using this insight, we develop new invertible transformations named convolution exponentials and graph convolution exponentials, which retain the equivariance of their underlying transformations. In addition, we generalize Sylvester Flows and propose Convolutional Sylvester Flows which are based on the generalization and the convolution exponential as basis change. Empirically, we show that the convolution exponential outperforms other linear transformations in generative flows on CIFAR10 and the graph convolution exponential improves the performance of graph normalizing flows. In addition, we show that Convolutional Sylvester Flows improve performance over residual flows as a generative flow model measured in log-likelihood.
- Europe > Netherlands > North Holland > Amsterdam (0.05)
- North America > Canada > British Columbia > Metro Vancouver Regional District > Vancouver (0.04)
- North America > Canada > British Columbia > Metro Vancouver Regional District > Vancouver (0.14)
- Europe > Netherlands > North Holland > Amsterdam (0.05)
- North America > United States > California > Los Angeles County > Long Beach (0.04)
- (5 more...)
The Convolution Exponential and Generalized Sylvester Flows
This paper introduces a new method to build linear flows, by taking the exponential of a linear transformation. This linear transformation does not need to be invertible itself, and the exponential has the following desirable properties: it is guaranteed to be invertible, its inverse is straightforward to compute and the log Jacobian determinant is equal to the trace of the linear transformation. An important insight is that the exponential can be computed implicitly, which allows the use of convolutional layers. Using this insight, we develop new invertible transformations named convolution exponentials and graph convolution exponentials, which retain the equivariance of their underlying transformations. In addition, we generalize Sylvester Flows and propose Convolutional Sylvester Flows which are based on the generalization and the convolution exponential as basis change.
The Convolution Exponential and Generalized Sylvester Flows
Hoogeboom, Emiel, Satorras, Victor Garcia, Tomczak, Jakub M., Welling, Max
This paper introduces a new method to build linear flows, by taking the exponential of a linear transformation. This linear transformation does not need to be invertible itself, and the exponential has the following desirable properties: it is guaranteed to be invertible, its inverse is straightforward to compute and the log Jacobian determinant is equal to the trace of the linear transformation. An important insight is that the exponential can be computed implicitly, which allows the use of convolutional layers. Using this insight, we develop new invertible transformations named convolution exponentials and graph convolution exponentials, which retain the equivariance of their underlying transformations. In addition, we generalize Sylvester Flows and propose Convolutional Sylvester Flows which are based on the generalization and the convolution exponential as basis change. Empirically, we show that the convolution exponential outperforms other linear transformations in generative flows on CIFAR10 and the graph convolution exponential improves the performance of graph normalizing flows. In addition, we show that Convolutional Sylvester Flows improve performance over residual flows as a generative flow model measured in log-likelihood.
- North America > Canada > British Columbia > Metro Vancouver Regional District > Vancouver (0.14)
- Europe > Netherlands > North Holland > Amsterdam (0.04)
- North America > United States > California > Los Angeles County > Long Beach (0.04)
- (5 more...)
The Expressive Power of a Class of Normalizing Flow Models
Kong, Zhifeng, Chaudhuri, Kamalika
Normalizing flows have received a great deal of recent attention as they allow flexible generative modeling as well as easy likelihood computation. While a wide variety of flow models have been proposed, there is little formal understanding of the representation power of these models. In this work, we study some basic normalizing flows and rigorously establish bounds on their expressive power. Our results indicate that while these flows are highly expressive in one dimension, in higher dimensions their representation power may be limited, especially when the flows have moderate depth.
- North America > United States > California > San Diego County > San Diego (0.04)
- Europe > Italy (0.04)
Sylvester Normalizing Flows for Variational Inference
Berg, Rianne van den, Hasenclever, Leonard, Tomczak, Jakub M., Welling, Max
Variational inference relies on flexible approximate posterior distributions. Normalizing flows provide a general recipe to construct flexible variational posteriors. We introduce Sylvester normalizing flows, which can be seen as a generalization of planar flows. Sylvester normalizing flows remove the well-known single-unit bottleneck from planar flows, making a single transformation much more flexible. We compare the performance of Sylvester normalizing flows against planar flows and inverse autoregressive flows and demonstrate that they compare favorably on several datasets.
- Europe > Netherlands > North Holland > Amsterdam (0.04)
- Asia > Middle East > Jordan (0.04)
- Europe > United Kingdom > England > Oxfordshire > Oxford (0.04)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Uncertainty (0.94)
- Information Technology > Artificial Intelligence > Machine Learning > Statistical Learning (0.68)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.68)